-
Notifications
You must be signed in to change notification settings - Fork 4.8k
Temporarily disable flaky volume test #21865
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Temporarily disable flaky volume test #21865
Conversation
`[sig-storage] Volume limits should verify that all nodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]`
|
Is this flaky, or is it just dead? I may have hit this on every run since the rebase ;). |
|
unit: although I don't see how that would be due to f56f493, so /retest |
|
/lgtm for good measure ;). |
|
[APPROVALNOTIFIER] This PR is APPROVED This pull-request has been approved by: smarterclayton, wking The full list of commands accepted by this bot can be found here. The pull request process is described here DetailsNeeds approval from an approver in each of these files:
Approvers can indicate their approval by writing |
|
It succeeds extremely rarely. It just so happened that was the two prior runs to the rebase, so we thought it was gravy. |
|
/test e2e-aws |
|
I'm waiting for a green e2e-aws and then I'm going to force merge |
|
/retest |
|
This requires kublet-1.12 because it is a new feature that depends on volume plugin being registered on the node. So far flakes I have investigated related to this are still using 1.11 kubelet. But I am still looking. |
/test e2e-aws-builds |
|
/test e2e-aws |
/test e2e-aws |
|
/retest Please review the full test history for this PR and help us cut down flakes. |
|
/retest |
Looking at origin's e2e-aws history, we haven't had anything pass in 11+ hours. So if you don't mind waiting it out while whatever the current batch job is fails, maybe this will just get merged without having to force it. Or maybe Tide will decide it needs to retest it in zounds of possible batch combinations, I dunno ;). |
|
It's possible this run is dying with: although I don't know why that would still be running after an hour. You may want to bump your commit timestamp or something to get a fresh, new namespace. |
|
I’m more concerned that teardown isn’t running. Is this the installer backgrounding? |
|
/test e2e-aws |
|
No teardown logs from this last run, but here's a normal teardown from a recent installer-PR failure: |
|
My guess is the issue was job 2885 or a similar run. My guess is artifact gathering was slow "no cluster", and we reaped teardown before it completed. But I don't seee any |
|
I’ve noticed some of that in jobs today. |
|
/retest |
|
We should short-circuit artifact gathering when Terraform fails. Have the installer exit 2? Grep the logs? |
|
Ah, or gate on some really-basic API call succeeding. I can work that up tomorrow if you don't beat me to it ;). |
|
I’ll probably be looking at other things so don’t worry about me also looking at it :) |
|
What the fork. |
|
One risk is that an api failure flake could result in no logs. We should definitely consider something like queue terminating early if enough sequential things aren’t gathered. |
|
@smarterclayton: The following test failed, say
Full PR test history. Your PR dashboard. Please help us cut down on flakes by linking to an open issue when you hit one in your PR. DetailsInstructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. I understand the commands that are listed here. |
| `openshift mongodb replication creating from a template`, // flaking on deployment | ||
| `should use be able to process many pods and reuse local volumes`, // https://bugzilla.redhat.com/show_bug.cgi?id=1635893 | ||
|
|
||
| `[sig-storage] Volume limits should verify that all nodes have volume limits`, // flaking due to a kubelet issue |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I saw this again in a job kicked off after the merge. Maybe the leading [sig-storage] here is a problem? The entries above don't seem to have those.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yeah, it needed to be regex quoted. Manually check and merged a follow up.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Cross-linking #21867.
[sig-storage] Volume limits should verify that all nodes have volume limits [Suite:openshift/conformance/parallel] [Suite:k8s]Introduced in rebase, https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/21860/pull-ci-openshift-origin-master-e2e-aws/2856/
@wongma7 @gnufied fyi